Goto

Collaborating Authors

 mayo clinic


Longitudinal and Multimodal Recording System to Capture Real-World Patient-Clinician Conversations for AI and Encounter Research: Protocol

Zahidy, Misk Al, Maldonado, Kerly Guevara, Andrango, Luis Vilatuna, Proano, Ana Cristina, Claros, Ana Gabriela, Jimenez, Maria Lizarazo, Toro-Tobon, David, Montori, Victor M., Ponce-Ponte, Oscar J., Brito, Juan P.

arXiv.org Artificial Intelligence

The promise of AI in medicine depends on learning from data that reflect what matters to patients and clinicians. Most existing models are trained on electronic health records (EHRs), which capture biological measures but rarely patient-clinician interactions. These relationships, central to care, unfold across voice, text, and video, yet remain absent from datasets. As a result, AI systems trained solely on EHRs risk perpetuating a narrow biomedical view of medicine and overlooking the lived exchanges that define clinical encounters. Our objective is to design, implement, and evaluate the feasibility of a longitudinal, multimodal system for capturing patient-clinician encounters, linking 360 degree video/audio recordings with surveys and EHR data to create a dataset for AI research. This single site study is in an academic outpatient endocrinology clinic at Mayo Clinic. Adult patients with in-person visits to participating clinicians are invited to enroll. Encounters are recorded with a 360 degree video camera. After each visit, patients complete a survey on empathy, satisfaction, pace, and treatment burden. Demographic and clinical data are extracted from the EHR. Feasibility is assessed using five endpoints: clinician consent, patient consent, recording success, survey completion, and data linkage across modalities. Recruitment began in January 2025. By August 2025, 35 of 36 eligible clinicians (97%) and 212 of 281 approached patients (75%) had consented. Of consented encounters, 162 (76%) had complete recordings and 204 (96%) completed the survey. This study aims to demonstrate the feasibility of a replicable framework for capturing the multimodal dynamics of patient-clinician encounters. By detailing workflows, endpoints, and ethical safeguards, it provides a template for longitudinal datasets and lays the foundation for AI models that incorporate the complexity of care.


A Rule Based Solution to Co-reference Resolution in Clinical Text

Chen, Ping, Hinote, David, Chen, Guoqing

arXiv.org Artificial Intelligence

Objective: The aim of this study was to build an effective co-reference resolution system tailored for the biomedical domain. Materials and Methods: Experiment materials used in this study is provided by the 2011 i2b2 Natural Language Processing Challenge. The 2011 i2b2 challenge involves coreference resolution in medical documents. Concept mentions have been annotated in clinical texts, and the mentions that co-refer in each document are to be linked by coreference chains. Normally, there are two ways of constructing a system to automatically discover co-referent links. One is to manually build rules for co-reference resolution, and the other category of approaches is to use machine learning systems to learn automatically from training datasets and then perform the resolution task on testing datasets. Results: Experiments show the existing co-reference resolution systems are able to find some of the co-referent links, and our rule based system performs well finding the majority of the co-referent links. Our system achieved 89.6% overall performance on multiple medical datasets. Conclusion: The experiment results show that manually crafted rules based on observation of training data is a valid way to accomplish high performance in this coreference resolution task for the critical biomedical domain.


Why it's so hard to use AI to diagnose cancer

MIT Technology Review

In theory, artificial intelligence should be great at helping out. "Our job is pattern recognition," says Andrew Norgan, a pathologist and medical director of the Mayo Clinic's digital pathology platform. "We look at the slide and we gather pieces of information that have been proven to be important." Visual analysis is something that AI has gotten quite good at since the first image recognition models began taking off nearly 15 years ago. Even though no model will be perfect, you can imagine a powerful algorithm someday catching something that a human pathologist missed, or at least speeding up the process of getting a diagnosis.


Atlas: A Novel Pathology Foundation Model by Mayo Clinic, Charit\'e, and Aignostics

Alber, Maximilian, Tietz, Stephan, Dippel, Jonas, Milbich, Timo, Lesort, Timothée, Korfiatis, Panos, Krügener, Moritz, Cancer, Beatriz Perez, Shah, Neelay, Möllers, Alexander, Seegerer, Philipp, Carpen-Amarie, Alexandra, Standvoss, Kai, Dernbach, Gabriel, de Jong, Edwin, Schallenberg, Simon, Kunft, Andreas, von Ankershoffen, Helmut Hoffer, Schaeferle, Gavin, Duffy, Patrick, Redlon, Matt, Jurmeister, Philipp, Horst, David, Ruff, Lukas, Müller, Klaus-Robert, Klauschen, Frederick, Norgan, Andrew

arXiv.org Artificial Intelligence

Recent advances in digital pathology have demonstrated the effectiveness of foundation models across diverse applications. In this report, we present Atlas, a novel vision foundation model based on the RudolfV approach. Our model was trained on a dataset comprising 1.2 million histopathology whole slide images, collected from two medical institutions: Mayo Clinic and Charit\'e - Universt\"atsmedizin Berlin. Comprehensive evaluations show that Atlas achieves state-of-the-art performance across twenty-one public benchmark datasets, even though it is neither the largest model by parameter count nor by training dataset size.


AI fast-tracks dementia diagnoses by tapping into 'hidden information' in brain waves

FOX News

As dementia becomes more widespread, Mayo Clinic researchers believe that artificial intelligence is the key to enabling earlier and faster diagnoses. By pairing AI and EEG (electroencephalogram) tests, the team at the Mayo Clinic Neurology AI Program (NAIP) in Rochester, Minnesota, was able to identify specific types of dementia sooner than they would have through human analysis. Based on these findings, EEGs could eventually provide a more accessible, less expensive and less invasive way to assess brain health earlier, according to a hospital press release. The research was published last week in the journal Brain Communications. With an EEG, a technician attaches small metal electrodes to the patient's scalp, which measure electrical activity in the brain.


MedYOLO: A Medical Image Object Detection Framework

Sobek, Joseph, Inojosa, Jose R. Medina, Inojosa, Betsy J. Medina, Rassoulinejad-Mousavi, S. M., Conte, Gian Marco, Lopez-Jimenez, Francisco, Erickson, Bradley J.

arXiv.org Artificial Intelligence

Artificial intelligence-enhanced identification of organs, lesions, and other structures in medical imaging is typically done using convolutional neural networks (CNNs) designed to make voxel-accurate segmentations of the region of interest. However, the labels required to train these CNNs are time-consuming to generate and require attention from subject matter experts to ensure quality. For tasks where voxel-level precision is not required, object detection models offer a viable alternative that can reduce annotation effort. Despite this potential application, there are few options for general purpose object detection frameworks available for 3-D medical imaging. We report on MedYOLO, a 3-D object detection framework using the one-shot detection method of the YOLO family of models and designed for use with medical imaging. We tested this model on four different datasets: BRaTS, LIDC, an abdominal organ Computed Tomography (CT) dataset, and an ECG-gated heart CT dataset. We found our models achieve high performance on commonly present medium and large-sized structures such as the heart, liver, and pancreas even without hyperparameter tuning. However, the models struggle with very small or rarely present structures.


Deep Phenotyping of Non-Alcoholic Fatty Liver Disease Patients with Genetic Factors for Insights into the Complex Disease

Priya, Tahmina Sultana, Leng, Fan, Luehrs, Anthony C., Klee, Eric W., Allen, Alina M., Lazaridis, Konstantinos N., Danfeng, null, Yao, null, Tian, Shulan

arXiv.org Artificial Intelligence

Non-alcoholic fatty liver disease (NAFLD) is a prevalent chronic liver disorder characterized by the excessive accumulation of fat in the liver in individuals who do not consume significant amounts of alcohol, including risk factors like obesity, insulin resistance, type 2 diabetes, etc. We aim to identify subgroups of NAFLD patients based on demographic, clinical, and genetic characteristics for precision medicine. The genomic and phenotypic data (3,408 cases and 4,739 controls) for this study were gathered from participants in Mayo Clinic Tapestry Study (IRB#19-000001) and their electric health records, including their demographic, clinical, and comorbidity data, and the genotype information through whole exome sequencing performed at Helix using the Exome+$^\circledR$ Assay according to standard procedure (www$.$helix$.$com). Factors highly relevant to NAFLD were determined by the chi-square test and stepwise backward-forward regression model. Latent class analysis (LCA) was performed on NAFLD cases using significant indicator variables to identify subgroups. The optimal clustering revealed 5 latent subgroups from 2,013 NAFLD patients (mean age 60.6 years and 62.1% women), while a polygenic risk score based on 6 single-nucleotide polymorphism (SNP) variants and disease outcomes were used to analyze the subgroups. The groups are characterized by metabolic syndrome, obesity, different comorbidities, psychoneurological factors, and genetic factors. Odds ratios were utilized to compare the risk of complex diseases, such as fibrosis, cirrhosis, and hepatocellular carcinoma (HCC), as well as liver failure between the clusters. Cluster 2 has a significantly higher complex disease outcome compared to other clusters. Keywords: Fatty liver disease; Polygenic risk score; Precision medicine; Deep phenotyping; NAFLD comorbidities; Latent class analysis.


Mayo Clinic sees AI as 'transformative force' in health care, appoints Dr. Bhavik Patel as chief AI officer

FOX News

Fox News contributor Dr. Marc Siegel weighs in on how artificial intelligence can change the patient-doctor relationship on'America's Newsroom.' As artificial intelligence gains an ever-widening role in the medical field, the Mayo Clinic has recently appointed a new executive to lead the health system's efforts in that area. Radiologist Bhavik Patel, M.D., has been named chief artificial intelligence officer (CAIO) for Mayo Clinic Arizona. Before joining the clinic in 2021, Patel practiced at Duke University Medical Center and Stanford University Medical Center. Dr. Richard Gray, CEO of Mayo Clinic Arizona, announced the hire on LinkedIn, noting the organization has only "begun to scratch the surface of AI's potential in medicine."


What Happened to All Those Jobs ChatGPT Was Supposed to Nuke?

Slate

This article is from Big Technology, a newsletter by Alex Kantrowitz. As soon as artificial intelligence began to read, write, and code, all manner of professions were supposed to automate--fast. And yet, eight months after the release of ChatGPT--and several years since the advent of other A.I. business tools--the fallout's been muted. A.I. is being widely adopted, but the imagined mass firings haven't materialized. The United States is still effectively at full employment, with just 3.5 percent of the workforce unemployed. The usual narrative may say otherwise, but the path toward A.I.–driven mass unemployment isn't simple.


Google is testing its medical AI chatbot at the Mayo Clinic

Engadget

Google is already testing its Med-PaLM 2 AI chat technology at at the Mayo Clinic and other hospitals, The Wall Street Journal has reported. It's based on the company's PaLM 2 large language model (LLM) that underpins Bard, Google's ChatGPT rival -- and was launched just months ago at Google I/O. Unlike the base model, Med-PaLM-2 has been trained on questions and answer from medical licensing exams, along with a curated set of medical expert demonstrations. That gives it expertise in answering health-related questions, and it can also do labor-intensive tasks like summarizing documents and organizing research data, according to the report. During I/O, Google released a paper detailing its work on Med-PaLM2.